16 research outputs found
Accelerating and Compressing Deep Neural Networks for Massive MIMO CSI Feedback
The recent advances in machine learning and deep neural networks have made
them attractive candidates for wireless communications functions such as
channel estimation, decoding, and downlink channel state information (CSI)
compression. However, most of these neural networks are large and inefficient
making it a barrier for deployment in practical wireless systems that require
low-latency and low memory footprints for individual network functions. To
mitigate these limitations, we propose accelerated and compressed efficient
neural networks for massive MIMO CSI feedback. Specifically, we have thoroughly
investigated the adoption of network pruning, post-training dynamic range
quantization, and weight clustering to optimize CSI feedback compression for
massive MIMO systems. Furthermore, we have deployed the proposed model
compression techniques on commodity hardware and demonstrated that in order to
achieve inference gains, specialized libraries that accelerate computations for
sparse neural networks are required. Our findings indicate that there is
remarkable value in applying these model compression techniques and the
proposed joint pruning and quantization approach reduced model size by 86.5%
and inference time by 76.2% with minimal impact to model accuracy. These
compression methods are crucial to pave the way for practical adoption and
deployments of deep learning-based techniques in commercial wireless systems.Comment: IEEE ICC 2023 Conferenc
Using Early Exits for Fast Inference in Automatic Modulation Classification
Automatic modulation classification (AMC) plays a critical role in wireless
communications by autonomously classifying signals transmitted over the radio
spectrum. Deep learning (DL) techniques are increasingly being used for AMC due
to their ability to extract complex wireless signal features. However, DL
models are computationally intensive and incur high inference latencies. This
paper proposes the application of early exiting (EE) techniques for DL models
used for AMC to accelerate inference. We present and analyze four early exiting
architectures and a customized multi-branch training algorithm for this
problem. Through extensive experimentation, we show that signals with moderate
to high signal-to-noise ratios (SNRs) are easier to classify, do not require
deep architectures, and can therefore leverage the proposed EE architectures.
Our experimental results demonstrate that EE techniques can significantly
reduce the inference speed of deep neural networks without sacrificing
classification accuracy. We also thoroughly study the trade-off between
classification accuracy and inference time when using these architectures. To
the best of our knowledge, this work represents the first attempt to apply
early exiting methods to AMC, providing a foundation for future research in
this area
How Does Forecasting Affect the Convergence of DRL Techniques in O-RAN Slicing?
The success of immersive applications such as virtual reality (VR) gaming and
metaverse services depends on low latency and reliable connectivity. To provide
seamless user experiences, the open radio access network (O-RAN) architecture
and 6G networks are expected to play a crucial role. RAN slicing, a critical
component of the O-RAN paradigm, enables network resources to be allocated
based on the needs of immersive services, creating multiple virtual networks on
a single physical infrastructure. In the O-RAN literature, deep reinforcement
learning (DRL) algorithms are commonly used to optimize resource allocation.
However, the practical adoption of DRL in live deployments has been sluggish.
This is primarily due to the slow convergence and performance instabilities
suffered by the DRL agents both upon initial deployment and when there are
significant changes in network conditions. In this paper, we investigate the
impact of time series forecasting of traffic demands on the convergence of the
DRL-based slicing agents. For that, we conduct an exhaustive experiment that
supports multiple services including real VR gaming traffic. We then propose a
novel forecasting-aided DRL approach and its respective O-RAN practical
deployment workflow to enhance DRL convergence. Our approach shows up to 22.8%,
86.3%, and 300% improvements in the average initial reward value, convergence
rate, and number of converged scenarios respectively, enhancing the
generalizability of the DRL agents compared with the implemented baselines. The
results also indicate that our approach is robust against forecasting errors
and that forecasting models do not have to be ideal.Comment: This article has been accepted for presentation in IEEE GLOBECOM 202
Lookback scheduling for long-term quality-of-service over multiple cells
Abstract-In current cellular networks, schedulers allocate wireless channel resources to users based on short-term moving averages of the channel gain and of the queuing state. Using only such short-term information, schedulers ignore the user's service history in previous cells and, thus, cannot meet long-term Quality of Service (QoS) guarantees when users traverse cells with varying load and capacity. We propose a new scheduling framework, which extends conventional short-term scheduling with long-term QoS information from previously traversed cells. We demonstrate our scheme for relevant channel-aware as well as for channel and queue-aware schedulers. Our simulation results show high gains in long-term QoS while the average throughput of the network increases. Therefore, the proposed scheduling approach improves subscriber satisfaction while increasing operational efficiency
Segmented Learning for Class-of-Service Network Traffic Classification
Class-of-service (CoS) network traffic classification (NTC) classifies a
group of similar traffic applications. The CoS classification is advantageous
in resource scheduling for Internet service providers and avoids the necessity
of remodelling. Our goal is to find a robust, lightweight, and fast-converging
CoS classifier that uses fewer data in modelling and does not require
specialized tools in feature extraction. The commonality of statistical
features among the network flow segments motivates us to propose novel
segmented learning that includes essential vector representation and a
simple-segment method of classification. We represent the segmented traffic in
the vector form using the EVR. Then, the segmented traffic is modelled for
classification using random forest. Our solution's success relies on finding
the optimal segment size and a minimum number of segments required in
modelling. The solution is validated on multiple datasets for various CoS
services, including virtual reality (VR). Significant findings of the research
work are i) Synchronous services that require acknowledgment and request to
continue communication are classified with 99% accuracy, ii) Initial 1,000
packets in any session are good enough to model a CoS traffic for promising
results, and we therefore can quickly deploy a CoS classifier, and iii) Test
results remain consistent even when trained on one dataset and tested on a
different dataset. In summary, our solution is the first to propose
segmentation learning NTC that uses fewer features to classify most CoS traffic
with an accuracy of 99%. The implementation of our solution is available on
GitHub.Comment: The paper is accepted to be appeared in IEEE GLOBECOM 202